Selecting Input Variables Using Mutual Informationand Nonparametric Density

نویسندگان

  • Brian V. Bonnlander
  • Andreas S. Weigend
چکیده

In learning problems where a connectionist network is trained with a nite sized training set, better generalization performance is often obtained when unneeded weights in the network are eliminated. One source of unneeded weights comes from the inclusion of input variables that provide little information about the output variables. We propose a method for identifying and eliminating these input variables. The method rst determines the relationship between input and output variables using nonparametric density estimation and then measures the relevance of input variables using the information theoretic concept of mutual information. We present results from our method on a simple toy problem and a nonlinear time series.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

ارائه مدلی غیرپارامتریک با استفاده از تکنیک k- نزدیک‌ترین همسایه در برآورد جرم مخصوص ظاهری خاک

Soil bulk density measurements are often required as an input parameter for models that predict soil processes. Nonparametric approaches are being used in various fields to estimate continuous variables. One type of the nonparametric lazy learning algorithms, a k-nearest neighbor (k-NN) algorithm was introduced and tested to estimate soil bulk density from other soil properties, including soil ...

متن کامل

Statistical Topology Using the Nonparametric Density Estimation and Bootstrap Algorithm

This paper presents approximate confidence intervals for each function of parameters in a Banach space based on a bootstrap algorithm. We apply kernel density approach to estimate the persistence landscape. In addition, we evaluate the quality distribution function estimator of random variables using integrated mean square error (IMSE). The results of simulation studies show a significant impro...

متن کامل

Estimation and Classification by Sigmoids based on Mutual Information

An estimate of the probability density function of a random vector is obtained by maximizing the mutual information between the input and the output of a feedforward network of sigmoidal units with respect to the input weights. Classification problems can be solved by selecting the class associated with the maximal estimated density. Newton's method, applied to an estimated density, yields a re...

متن کامل

Renyi's-entropy-based Approach for Selecting the Significant Input Variables for the Ecological data

Recently, data-driven approaches including machine-learning (ML) techniques have played a key role in the research on ecological data and models. One of the most important steps in the application of a ML technique is the selection of significant model input variables. Among ML methods, artificial neural networks and genetic algorithm are widely used for the sake of the above aim; however entro...

متن کامل

Application of Renyi Entropy and Mutual Informa-tion of Cauchy-Schwartz in Selecting Variables

This paper approaches the algorithm of selection of variables named MIFS-U and presents an alternative method for estimating entropy and mutual information, “measures” that constitute the base of this selection algorithm. This method has, for foundation, the Cauchy-Schwartz quadratic mutual information and the Rényi quadratic entropy, combined, in the case of continuous variables, with Parzen W...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1996